AI Prompt Action (Multi-LLM)

This action is suited to execute LLM-prompts from multiple AI Providers, including on-prem once based on Ollama.

Actually it supports the following LLM providers and models

Supported LLM providers and models

Cohere

  • command-r
  • command-r-plus

Google

  • gemini-pro

Anthropic

  • claude-3-5-sonnet

Groq

Grouped by "creator"

Meta (LLaMA)

  • llama-3.2-11b-vision-preview
  • llama-3.1-8b-instant
  • llama-3.2-3b-preview
  • llama-3.2-90b-vision-preview
  • llama-guard-3-8b
  • llama3-8b-8192
  • llama-3.2-1b-preview
  • llama-3.3-70b-versatile
  • llama3-70b-8192
  • llama-3.3-70b-specdec

DeepSeek / Alibaba Cloud

  • deepseek-r1-distill-qwen-32b
  • deepseek-r1-distill-llama-70b

Alibaba Cloud

  • qwen-2.5-32b

Google

  • gemma2-9b-it

SDAIA

  • allam-2-7b

Mistral AI

  • mixtral-8x7b-32768

How to configure and use the action

Drag & Drop the Action from the actions palette on the left to the stage:

Once on the stage you can configure your favourite LLM prompt using the action detail panel.

To properly use you Action you need to configure it in the Settings > Integrations section, provinding the corresponding LLM API-KEY.

E.g. for Google Gemini:

We hope you enjoy our new Action that will let you use your favourite LLM provider and models!

If you have questions about the AI Prompt Action or other Tiledesk features feel free to send us an email to support@tiledesk.com  or leave us a feedback